The search functionality is under construction.

Keyword Search Result

[Keyword] EDGE computing(37hit)

21-37hit(37hit)

  • Empirical Study of Low-Latency Network Model with Orchestrator in MEC Open Access

    Krittin INTHARAWIJITR  Katsuyoshi IIDA  Hiroyuki KOGA  Katsunori YAMAOKA  

     
    PAPER-Network

      Pubricized:
    2020/09/01
      Vol:
    E104-B No:3
      Page(s):
    229-239

    The Internet of Things (IoT) with its support for cyber-physical systems (CPS) will provide many latency-sensitive services that require very fast responses from network services. Mobile edge computing (MEC), one of the distributed computing models, is a promising component of the low-latency network architecture. In network architectures with MEC, mobile devices will offload heavy computing tasks to edge servers. There exist numbers of researches about low-latency network architecture with MEC. However, none of the existing researches simultaneously satisfy the followings: (1) guarantee the latency of computing tasks and (2) implement a real system. In this paper, we designed and implemented an MEC based network architecture that guarantees the latency of offloading tasks. More specifically, we first estimate the total latency including computing and communication ones at the centralized node called orchestrator. If the estimated value exceeds the latency requirement, the task will be rejected. We then evaluated its performance in terms of the blocking probability of the tasks. To analyze the results, we compared the performance between obtained from experiments and simulations. Based on the comparisons, we clarified that the computing latency estimation accuracy is a significant factor for this system.

  • Benchmarking Modern Edge Devices for AI Applications

    Pilsung KANG  Jongmin JO  

     
    PAPER-Computer System

      Pubricized:
    2020/12/08
      Vol:
    E104-D No:3
      Page(s):
    394-403

    AI (artificial intelligence) has grown at an overwhelming speed for the last decade, to the extent that it has become one of the mainstream tools that drive the advancements in science and technology. Meanwhile, the paradigm of edge computing has emerged as one of the foremost areas in which applications using the AI technology are being most actively researched, due to its potential benefits and impact on today's widespread networked computing environments. In this paper, we evaluate two major entry-level offerings in the state-of-the-art edge device technology, which highlight increased computing power and specialized hardware support for AI applications. We perform a set of deep learning benchmarks on the devices to measure their performance. By comparing the performance with other GPU (graphics processing unit) accelerated systems in different platforms, we assess the computational capability of the modern edge devices featuring a significant amount of hardware parallelism.

  • Load Balancing for Energy-Harvesting Mobile Edge Computing

    Ping ZHAO  Jiawei TAO  Abdul RAUF  Fengde JIA  Longting XU  

     
    LETTER-Mobile Information Network and Personal Communications

      Pubricized:
    2020/07/27
      Vol:
    E104-A No:1
      Page(s):
    336-342

    With the development of cloud computing, the Mobile Edge Computing has emerged and attracted widespread attentions. In this paper, we focus on the load balancing in MEC with energy harvesting. We first introduce the load balancing in MEC as a problem of minimizing both the energy consumption and queue redundancy. Thereafter, we adapt such a optimization problem to the Lyapunov algorithm and solve this optimization problem. Finally, extensive simulation results validate that the obtained strategy improves the capabilities of MEC systems.

  • Construction of an Efficient Divided/Distributed Neural Network Model Using Edge Computing

    Ryuta SHINGAI  Yuria HIRAGA  Hisakazu FUKUOKA  Takamasa MITANI  Takashi NAKADA  Yasuhiko NAKASHIMA  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2020/07/02
      Vol:
    E103-D No:10
      Page(s):
    2072-2082

    Modern deep learning has significantly improved performance and has been used in a wide variety of applications. Since the amount of computation required for the inference process of the neural network is large, it is processed not by the data acquisition location like a surveillance camera but by the server with abundant computing power installed in the data center. Edge computing is getting considerable attention to solve this problem. However, edge computing can provide limited computation resources. Therefore, we assumed a divided/distributed neural network model using both the edge device and the server. By processing part of the convolution layer on edge, the amount of communication becomes smaller than that of the sensor data. In this paper, we have evaluated AlexNet and the other eight models on the distributed environment and estimated FPS values with Wi-Fi, 3G, and 5G communication. To reduce communication costs, we also introduced the compression process before communication. This compression may degrade the object recognition accuracy. As necessary conditions, we set FPS to 30 or faster and object recognition accuracy to 69.7% or higher. This value is determined based on that of an approximation model that binarizes the activation of Neural Network. We constructed performance and energy models to find the optimal configuration that consumes minimum energy while satisfying the necessary conditions. Through the comprehensive evaluation, we found that the optimal configurations of all nine models. For small models, such as AlexNet, processing entire models in the edge was the best. On the other hand, for huge models, such as VGG16, processing entire models in the server was the best. For medium-size models, the distributed models were good candidates. We confirmed that our model found the most energy efficient configuration while satisfying FPS and accuracy requirements, and the distributed models successfully reduced the energy consumption up to 48.6%, and 6.6% on average. We also found that HEVC compression is important before transferring the input data or the feature data between the distributed inference processes.

  • Participating-Domain Segmentation Based Server Selection Scheme for Real-Time Interactive Communication Open Access

    Akio KAWABATA  Bijoy CHAND CHATTERJEE  Eiji OKI  

     
    PAPER-Network

      Pubricized:
    2020/01/17
      Vol:
    E103-B No:7
      Page(s):
    736-747

    This paper proposes an efficient server selection scheme in successive participation scenario with participating-domain segmentation. The scheme is utilized by distributed processing systems for real-time interactive communication to suppress the communication latency of a wide-area network. In the proposed scheme, users participate for server selection one after another. The proposed scheme determines a recommended server, and a new user selects the recommended server first. Before each user participates, the recommended servers are determined assuming that users exist in the considered regions. A recommended server is determined for each divided region to minimize the latency. The new user selects the recommended available server, where the user is located. We formulate an integer linear programming problem to determine the recommended servers. Numerical results indicate that, at the cost additional computation, the proposed scheme offers smaller latency than the conventional scheme. We investigate different policies to divide the users' participation for the recommended server finding process in the proposed scheme.

  • Auction-Based Resource Allocation for Mobile Edge Computing Networks

    Ben LIU  Ding XU  

     
    LETTER-Communication Theory and Signals

      Vol:
    E103-A No:4
      Page(s):
    718-722

    Mobile edge computing (MEC) is a new computing paradigm, which provides computing support for resource-constrained user equipments (UEs). In this letter, we design an effective incentive framework to encourage MEC operators to provide computing service for UEs. The problem of jointly allocating communication and computing resources to maximize the revenue of MEC operators is studied. Based on auction theory, we design a multi-round iterative auction (MRIA) algorithm to solve the problem. Extensive simulations have been conducted to evaluate the performance of the proposed algorithm and it is shown that the proposed algorithm can significantly improve the overall revenue of MEC operators.

  • Essential Roles, Challenges and Development of Embedded MCU Micro-Systems to Innovate Edge Computing for the IoT/AI Age Open Access

    Takashi KONO  Yasuhiko TAITO  Hideto HIDAKA  

     
    INVITED PAPER-Integrated Electronics

      Vol:
    E103-C No:4
      Page(s):
    132-143

    Embedded system approaches to edge computing in IoT implementations are proposed and discussed. Rationales of edge computing and essential core capabilities for IoT data supply innovation are identified. Then, innovative roles and development of MCU and embedded flash memory are illustrated by technology and applications, expanding from CPS to big-data and nomadic/autonomous elements of IoT requirements. Conclusively, a technology roadmap construction specific to IoT is proposed.

  • On the Design and Implementation of IP-over-P2P Overlay Virtual Private Networks Open Access

    Kensworth SUBRATIE  Saumitra ADITYA  Vahid DANESHMAND  Kohei ICHIKAWA  Renato FIGUEIREDO  

     
    INVITED PAPER-Network

      Pubricized:
    2019/08/05
      Vol:
    E103-B No:1
      Page(s):
    2-10

    The success and scale of the Internet and its protocol IP has spurred emergent distributed technologies such as fog/edge computing and new application models based on distributed containerized microservices. The Internet of Things and Connected Communities are poised to build on these technologies and models and to benefit from the ability to communicate in a peer-to-peer (P2P) fashion. Ubiquitous sensing, actuating and computing implies a scale that breaks the centralized cloud computing model. Challenges stemming from limited IPv4 public addresses, the need for transport layer authentication, confidentiality and integrity become a burden on developing new middleware and applications designed for the network's edge. One approach - not reliant on the slow adoption of IPv6 - is the use of virtualized overlay networks, which abstract the complexities of the underlying heterogeneous networks that span the components of distributed fog applications and middleware. This paper describes the evolution of the design and implementation of IP-over-P2P (IPOP) - from its purist P2P inception, to a pragmatic hybrid model which is influenced by and incorporates standards. The hybrid client-server/P2P approach allows IPOP to leverage existing robust and mature cloud infrastructure, while still providing the characteristics needed at the edge. IPOP is networking cyber infrastructure that presents an overlay virtual private network which self-organizes with dynamic membership of peer nodes into a scalable structure. IPOP is resilient to partitioning, supports redundant paths within its fabric, and provides software defined programming of switching rules to utilize these properties of its topology.

  • Distributed Key-Value Storage for Edge Computing and Its Explicit Data Distribution Method

    Takehiro NAGATO  Takumi TSUTANO  Tomio KAMADA  Yumi TAKAKI  Chikara OHTA  

     
    PAPER-Network

      Pubricized:
    2019/08/05
      Vol:
    E103-B No:1
      Page(s):
    20-31

    In this article, we propose a data framework for edge computing that allows developers to easily attain efficient data transfer between mobile devices or users. We propose a distributed key-value storage platform for edge computing and its explicit data distribution management method that follows the publish/subscribe relationships specific to applications. In this platform, edge servers organize the distributed key-value storage in a uniform namespace. To enable fast data access to a record in edge computing, the allocation strategy of the record and its cache on the edge servers is important. Our platform offers distributed objects that can dynamically change their home server and allocate cache objects proactively following user-defined rules. A rule is defined in a declarative manner and specifies where to place cache objects depending on the status of the target record and its associated records. The system can reflect record modification to the cached records immediately. We also integrate a push notification system using WebSocket to notify events on a specified table. We introduce a messaging service application between mobile appliances and several other applications to show how cache rules apply to them. We evaluate the performance of our system using some sample applications.

  • Interworking Layer of Distributed MQTT Brokers

    Ryohei BANNO  Jingyu SUN  Susumu TAKEUCHI  Kazuyuki SHUDO  

     
    PAPER-Information Network

      Pubricized:
    2019/07/30
      Vol:
    E102-D No:12
      Page(s):
    2281-2294

    MQTT is one of the promising protocols for various data exchange in IoT environments. Typically, those environments have a characteristic called “edge-heavy”, which means that things at the network edge generate a massive volume of data with high locality. For handling such edge-heavy data, an architecture of placing multiple MQTT brokers at the network edges and making them cooperate with each other is quite effective. It can provide higher throughput and lower latency, as well as reducing consumption of cloud resources. However, under this kind of architecture, heterogeneity could be a vital issue. Namely, an appropriate product of MQTT broker could vary according to the different environment of each network edge, even though different products are hard to cooperate due to the MQTT specification providing no interoperability between brokers. In this paper, we propose Interworking Layer of Distributed MQTT brokers (ILDM), which enables arbitrary kinds of MQTT brokers to cooperate with each other. ILDM, designed as a generic mechanism independent of any specific cooperation algorithm, provides APIs to facilitate development of a variety of algorithms. By using the APIs, we also present two basic cooperation algorithms. To evaluate the usefulness of ILDM, we introduce a benchmark system which can be used for both a single broker and multiple brokers. Experimental results show that the throughput of five brokers running together by ILDM is improved 4.3 times at maximum than that of a single broker.

  • Simulation Study of Low-Latency Network Model with Orchestrator in MEC Open Access

    Krittin INTHARAWIJITR  Katsuyoshi IIDA  Hiroyuki KOGA  Katsunori YAMAOKA  

     
    PAPER-Network

      Pubricized:
    2019/05/16
      Vol:
    E102-B No:11
      Page(s):
    2139-2150

    Most of latency-sensitive mobile applications depend on computational resources provided by a cloud computing service. The problem of relying on cloud computing is that, sometimes, the physical locations of cloud servers are distant from mobile users and the communication latency is long. As a result, the concept of distributed cloud service, called mobile edge computing (MEC), is being introduced in the 5G network. However, MEC can reduce only the communication latency. The computing latency in MEC must also be considered to satisfy the required total latency of services. In this research, we study the impact of both latencies in MEC architecture with regard to latency-sensitive services. We also consider a centralized model, in which we use a controller to manage flows between users and mobile edge resources to analyze MEC in a practical architecture. Simulations show that the interval and controller latency trigger some blocking and error in the system. However, the permissive system which relaxes latency constraints and chooses an edge server by the lowest total latency can improve the system performance impressively.

  • Assessing Lightweight Virtualization for Security-as-a-Service at the Network Edge Open Access

    Abderrahmane BOUDI  Ivan FARRIS  Miloud BAGAA  Tarik TALEB  

     
    INVITED PAPER

      Pubricized:
    2018/11/22
      Vol:
    E102-B No:5
      Page(s):
    970-977

    Accounting for the exponential increase in security threats, the development of new defense strategies for pervasive environments is acquiring an ever-growing importance. The expected avalanche of heterogeneous IoT devices which will populate our industrial factories and smart houses will increase the complexity of managing security requirements in a comprehensive way. To this aim, cloud-based security services are gaining notable impetus to provide security mechanisms according to Security-as-a-Service (SECaaS) model. However, the deployment of security applications in remote cloud data-centers can introduce several drawbacks in terms of traffic overhead and latency increase. To cope with this, Edge Computing can provide remarkable advantages avoiding long routing detours. On the other hand, the limited capabilities of edge node introduce potential constraints in the overall management. This paper focuses on the provisioning of virtualized security services in resource-constrained edge nodes by leveraging lightweight virtualization technologies. Our analysis aims at shedding light on the feasibility of container-based security solutions, thus providing useful guidelines towards the orchestration of security at the edge. Our experiments show that the overhead introduced by the containerization is very light.

  • Mobile Network Architectures and Context-Aware Network Control Technology in the IoT Era Open Access

    Takanori IWAI  Daichi KOMINAMI  Masayuki MURATA  Ryogo KUBO  Kozo SATODA  

     
    INVITED PAPER

      Pubricized:
    2018/04/13
      Vol:
    E101-B No:10
      Page(s):
    2083-2093

    As IoT services become more popular, mobile networks will have to accommodate a wide variety of devices that have different requirements such as different bandwidth limitations and latencies. This paper describes edge distributed mobile network architectures for the IoT era based on dedicated network technology and multi-access edge computing technology, which have been discussed in 3GPP and ETSI. Furthermore, it describes two context-aware control methods that will make mobile networks on the network architecture more efficient, reliable, and real-time: autonomous and distributed mobility management and bandwidth-guaranteed transmission rate control in a networked control system.

  • Design and Implementation of Deep Neural Network for Edge Computing

    Junyang ZHANG  Yang GUO  Xiao HU  Rongzhen LI  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2018/05/02
      Vol:
    E101-D No:8
      Page(s):
    1982-1996

    In recent years, deep learning based image recognition, speech recognition, text translation and other related applications have brought great convenience to people's lives. With the advent of the era of internet of everything, how to run a computationally intensive deep learning algorithm on a limited resources edge device is a major challenge. For an edge oriented computing vector processor, combined with a specific neural network model, a new data layout method for putting the input feature maps in DDR, rearrangement of the convolutional kernel parameters in the nuclear memory bank is proposed. Aiming at the difficulty of parallelism of two-dimensional matrix convolution, a method of parallelizing the matrix convolution calculation in the third dimension is proposed, by setting the vector register with zero as the initial value of the max pooling to fuse the rectified linear unit (ReLU) activation function and pooling operations to reduce the repeated access to intermediate data. On the basis of single core implementation, a multi-core implementation scheme of Inception structure is proposed. Finally, based on the proposed vectorization method, we realize five kinds of neural network models, namely, AlexNet, VGG16, VGG19, GoogLeNet, ResNet18, and performance statistics and analysis based on CPU, gtx1080TI and FT2000 are presented. Experimental results show that the vector processor has better computing advantages than CPU and GPU, and can calculate large-scale neural network model in real time.

  • Mobile Edge Computing Empowers Internet of Things Open Access

    Nirwan ANSARI  Xiang SUN  

     
    INVITED PAPER

      Pubricized:
    2017/09/19
      Vol:
    E101-B No:3
      Page(s):
    604-619

    In this paper, we propose a Mobile Edge Internet of Things (MEIoT) architecture by leveraging the fiber-wireless access technology, the cloudlet concept, and the software defined networking framework. The MEIoT architecture brings computing and storage resources close to Internet of Things (IoT) devices in order to speed up IoT data sharing and analytics. Specifically, the IoT devices (belonging to the same user) are associated to a specific proxy Virtual Machine (VM) in the nearby cloudlet. The proxy VM stores and analyzes the IoT data (generated by its IoT devices) in real-time. Moreover, we introduce the semantic and social IoT technology in the context of MEIoT to solve the interoperability and inefficient access control problem in the IoT system. In addition, we propose two dynamic proxy VM migration methods to minimize the end-to-end delay between proxy VMs and their IoT devices and to minimize the total on-grid energy consumption of the cloudlets, respectively. Performance of the proposed methods is validated via extensive simulations.

  • Development of Wireless Access and Flexible Networking Technologies for 5G Cellular Systems Open Access

    Seiichi SAMPEI  

     
    INVITED PAPER-Wireless Communication Technologies

      Pubricized:
    2017/02/08
      Vol:
    E100-B No:8
      Page(s):
    1174-1180

    This paper discusses key technologies specific for fifth generation (5G) cellular systems which are expected to connect internet of things (IoT) based vertical sectors. Because services for 5G will be expanded drastically, from information transfer services to mission critical and massive connection IoT connection services for vertical sectors, and requirement for cellular systems becomes quite different compared to that of fourth generation (4G) systems, after explanation for the service and technical trends for 5G, key wireless access technologies will be discussed, especially, from the view point of what is new and how import. In addition to the introduction of new technologies for wireless access, flexibility of networking is also discussed because it can cope with QoS support services, especially to cope with end-to-end latency constraint conditions. Therefore, this paper also discuss flexible network configuration using mobile edge computing (MEC) based on software defined network (SDN) and network slicing.

  • Simulation Study of Low Latency Network Architecture Using Mobile Edge Computing

    Krittin INTHARAWIJITR  Katsuyoshi IIDA  Hiroyuki KOGA  

     
    PAPER

      Pubricized:
    2017/02/08
      Vol:
    E100-D No:5
      Page(s):
    963-972

    Attaining extremely low latency service in 5G cellular networks is an important challenge in the communication research field. A higher QoS in the next-generation network could enable several unprecedented services, such as Tactile Internet, Augmented Reality, and Virtual Reality. However, these services will all need support from powerful computational resources provided through cloud computing. Unfortunately, the geolocation of cloud data centers could be insufficient to satisfy the latency aimed for in 5G networks. The physical distance between servers and users will sometimes be too great to enable quick reaction within the service time boundary. The problem of long latency resulting from long communication distances can be solved by Mobile Edge Computing (MEC), though, which places many servers along the edges of networks. MEC can provide shorter communication latency, but total latency consists of both the transmission and the processing times. Always selecting the closest edge server will lead to a longer computing latency in many cases, especially when there is a mass of users around particular edge servers. Therefore, the research studies the effects of both latencies. The communication latency is represented by hop count, and the computation latency is modeled by processor sharing (PS). An optimization model and selection policies are also proposed. Quantitative evaluations using simulations show that selecting a server according to the lowest total latency leads to the best performance, and permitting an over-latency barrier would further improve results.

21-37hit(37hit)